# Mathematical Programming Enhancement
Deepseek R1 0528
MIT
The DeepSeek R1 model has undergone a minor version upgrade, with the current version being DeepSeek-R1-0528. In the latest update, DeepSeek R1 significantly enhanced its deep reasoning and inference capabilities by increasing computational resource investment and introducing algorithmic optimization mechanisms during the post-training phase.
Large Language Model
Transformers

D
deepseek-ai
4,556
1,249
Sarvam M
Apache-2.0
Sarvam-M is a multilingual, hybrid-reasoning, pure-text language model based on Mistral-Small, specifically optimized for Indian languages and English, featuring exceptional reasoning capabilities and cultural adaptability.
Large Language Model
Transformers Supports Multiple Languages

S
sarvamai
1,824
159
Qwen3 4B GGUF
Apache-2.0
Qwen3 is the latest version of the Tongyi Qianwen series of large language models, offering a range of dense and mixture-of-experts (MoE) models. Based on large-scale training, Qwen3 has achieved breakthrough progress in reasoning, instruction following, agent capabilities, and multilingual support.
Large Language Model
Q
Qwen
4,225
6
Qwen3 14B GGUF
Qwen3-14B is a large language model developed by Qwen, supporting multiple languages and dialects, excelling in creative writing, role-playing, and multi-turn conversations.
Large Language Model
Q
lmstudio-community
56.39k
9
M1 32b
Apache-2.0
M1-32B is a 32-billion-parameter large language model fine-tuned from Qwen2.5-32B-Instruct, specifically optimized to enhance reasoning, discussion, and decision-making capabilities in multi-agent systems.
Large Language Model
Transformers Supports Multiple Languages

M
Can111
179
3
35b Beta Long
A 35-billion-parameter multilingual large language model fine-tuned based on the CohereForAI/c4ai-command-r-v01 architecture, specializing in long-context processing and fact-based data generation
Large Language Model
Transformers Supports Multiple Languages

3
CausalLM
79
65
Beyonder 4x7B V2
Other
Beyonder-4x7B-v2 is a large language model based on the Mixture of Experts (MoE) architecture, consisting of 4 expert modules, each specializing in different domains such as dialogue, programming, creative writing, and mathematical reasoning.
Large Language Model
Transformers

B
mlabonne
758
130
Featured Recommended AI Models